Training Deep Models for Semantic Segmentation

This topic provides information for training deep models for semantic segmentation. Semantic segmentation is the task of classifying each pixel in an image from a predefined set of classes that are 'semantically interpretable' and correspond to real-world categories. This is also known as dense prediction because it predicts the meaning of each pixel.

In many cases, deep learning has surpassed other approaches, such as thresholding, K-means clustering, classic machine learning, and others, for semantic segmentation. In addition, applying a model on the same kind of data lets you benefit from consistent and repeatable segmentation results that are not influenced by any operator bias.

Refer to the recorded in-depth lessons and case studies in Training Deep Models for Image Segmentation to view overviews of training deep models for semantic segmentation. You also view this video and others about deep learning on our YouTube channel (https://www.youtube.com/channel/UCuFl2zHcyStR2RJpMXbi6ow).
Semantic segmentation is different from object detection (see Object Detection with YOLOv3) as it does not predict any bounding boxes around the objects. As such, different instances of the same object are not distinguished. For example, there could be multiple fibers in an image and all of them would have the same class.